Goto

Collaborating Authors

 channel prediction


A Wireless Foundation Model for Multi-Task Prediction

Sheng, Yucheng, Wang, Jiacheng, Zhou, Xingyu, Liang, Le, Ye, Hao, Jin, Shi, Li, Geoffrey Ye

arXiv.org Artificial Intelligence

--With the growing complexity and dynamics of the mobile communication networks, accurately predicting key system parameters, such as channel state information (CSI), user location, and network traffic, has become essential for a wide range of physical (PHY)-layer and medium access control (MAC)-layer tasks. Although traditional deep learning (DL)- based methods have been widely applied to such prediction tasks, they often struggle to generalize across different scenarios and tasks. In response, we propose a unified foundation model for multi-task prediction in wireless networks that supports diverse prediction intervals. The proposed model enforces univariate decomposition to unify heterogeneous tasks, encodes granularity for interval awareness, and uses a causal Transformer backbone for accurate predictions. Additionally, we introduce a patch masking strategy during training to support arbitrary input lengths. After trained on large-scale datasets, the proposed foundation model demonstrates strong generalization to unseen scenarios and achieves zero-shot performance on new tasks that surpass traditional full-shot baselines. HE advent of 6G communications [1] has made wireless systems more intricate, featuring ultra-dense deployments, diverse service demands, and highly dynamic environments. Efficient execution of physical (PHY) and medium access control (MAC)-layer tasks require accurate and timely knowledge of the surrounding communication environment. Key parameters of interest include channel state information (CSI) [2], user locations [3], mobile traffic at the base station (BS) [4], etc. However, these parameters fluctuate rapidly over time, making real-time estimation and feedback particularly challenging. As a result, accurately predicting these variables has become essential for enabling a wide range of downstream communication tasks.


Large Language Models for Wireless Communications: From Adaptation to Autonomy

Liang, Le, Ye, Hao, Sheng, Yucheng, Wang, Ouya, Wang, Jiacheng, Jin, Shi, Li, Geoffrey Ye

arXiv.org Artificial Intelligence

--The emergence of large language models (LLMs) has revolutionized artificial intelligence, offering unprecedented capabilities in reasoning, generalization, and zero-shot learning. These strengths open new frontiers in wireless communications, where increasing complexity and dynamics demand intelligent and adaptive solutions. This article explores the role of LLMs in transforming wireless systems across three key directions: adapting pretrained LLMs for core communication tasks, developing wireless-specific foundation models to balance versatility and efficiency, and enabling agentic LLMs with autonomous reasoning and coordination capabilities. We highlight recent advances, practical case studies, and the unique benefits of LLM-based approaches over traditional methods. Finally, we outline open challenges and research opportunities--including multimodal fusion, collaboration with lightweight models, and self-improving capabilities--charting a path toward intelligent, adaptive, and autonomous wireless networks of the future. The rapid advancement of large language models (LLMs) has transformed natural language processing, unlocking capabilities in reasoning, representation learning, and generalization from limited supervision. These models, built on transformer architectures and trained on large-scale text corpora, exhibit remarkable adaptability across tasks and domains. As such, their core strengths--sequence modeling, contextual understanding, and zero-shot inference--are increasingly being explored for applications far beyond language, to include robotics, software engineering, and, more recently, wireless communications. This article investigates how LLMs can be strategically repurposed to address key challenges in modern wireless networks, tracing a trajectory from task-specific model adaptation to the realization of autonomous, agent-driven communication systems. Next-generation wireless systems are characterized by growing complexity and variability.


Digital Twin Channel-Enabled Online Resource Allocation for 6G: Principle, Architecture and Application

Li, Tongjie, Zhang, Jianhua, Yu, Li, Zhang, Yuxiang, Cai, Yunlong, Xu, Fan, Liu, Guangyi

arXiv.org Artificial Intelligence

The emergence of sixth-generation (6G) networks is reshaping wireless communications to support mission-critical applications such as the Industrial Internet of Things (IIoT), autonomous driving, and smart manufacturing. Compared with 5G, 6G imposes significantly more stringent requirements on latency, reliability, adaptability, and end-to-end responsiveness [1, 2]. IIoT scenarios are particularly challenging due to the coexistence of complex radio propagation conditions and diverse service requirements. Dense deployments, metallic scatterers, and dynamic obstacles give rise to severe multipath fading, especially in high-frequency bands such as mmWave and terahertz, where signal stability is highly sensitive to physical structures [3, 4]. In parallel, service demands span multiple categories, such as periodic sensing, closed-loop control, event-triggered communication, and edge computing, each with distinct quality-of-service (QoS) requirements [5]. To address these multifaceted challenges, resource allocation mechanisms must be environment-aware, latency-sensitive, and capable of online adaptation across large-scale, dynamic deployments. Artificial intelligence (AI)-driven resource allocation has attracted growing interest due to its ability to learn underlying correlations from sensing data and historical records.


Large AI Model for Delay-Doppler Domain Channel Prediction in 6G OTFS-Based Vehicular Networks

Xue, Jianzhe, Yuan, Dongcheng, Ma, Zhanxi, Jiang, Tiankai, Sun, Yu, Zhou, Haibo, Shen, Xuemin

arXiv.org Artificial Intelligence

Channel prediction is crucial for high-mobility vehicular networks, as it enables the anticipation of future channel conditions and the proactive adjustment of communication strategies. However, achieving accurate vehicular channel prediction is challenging due to significant Doppler effects and rapid channel variations resulting from high-speed vehicle movement and complex propagation environments. In this paper, we propose a novel delay-Doppler (DD) domain channel prediction framework tailored for high-mobility vehicular networks. By transforming the channel representation into the DD domain, we obtain an intuitive, sparse, and stable depiction that closely aligns with the underlying physical propagation processes, effectively reducing the complex vehicular channel to a set of time-series parameters with enhanced predictability. Furthermore, we leverage the large artificial intelligence (AI) model to predict these DD-domain time-series parameters, capitalizing on their advanced ability to model temporal correlations. The zero-shot capability of the pre-trained large AI model facilitates accurate channel predictions without requiring task-specific training, while subsequent fine-tuning on specific vehicular channel data further improves prediction accuracy. Extensive simulation results demonstrate the effectiveness of our DD-domain channel prediction framework and the superior accuracy of the large AI model in predicting time-series channel parameters, thereby highlighting the potential of our approach for robust vehicular communication systems.


Edge Large AI Models: Revolutionizing 6G Networks

Wang, Zixin, Shi, Yuanming, Zhou, Yong, Zhu, Jingyang, Letaief, Khaled. B.

arXiv.org Artificial Intelligence

--Large artificial intelligence models (LAMs) possess human-like abilities to solve a wide range of real-world problems, exemplifying the potential of experts in various domains and modalities. By leveraging the communication and computation capabilities of geographically dispersed edge devices, edge LAM emerges as an enabling technology to empower the delivery of various real-time intelligent services in 6G. Unlike traditional edge artificial intelligence (AI) that primarily supports a single task using small models, edge LAM is featured by the need of the decomposition and distributed deployment of large models, and the ability to support highly generalized and diverse tasks. However, due to limited communication, computation, and storage resources over wireless networks, the vast number of trainable neurons and the substantial communication overhead pose a formidable hurdle to the practical deployment of edge LAMs. In this paper, we investigate the opportunities and challenges of edge LAMs from the perspectives of model decomposition and resource management. Specifically, we propose collaborative fine-tuning and full-parameter training frameworks, alongside a microservice-assisted inference architecture, to enhance the deployment of edge LAM over wireless networks. Additionally, we investigate the application of edge LAM in air-interface designs, focusing on channel prediction and beamforming. These innovative frameworks and applications offer valuable insights and solutions for advancing 6G technology. With the remarkable advancement in artificial intelligence (AI), large AI models (LAMs) now excel at performing real-world complex tasks.


WirelessGPT: A Generative Pre-trained Multi-task Learning Framework for Wireless Communication

Yang, Tingting, Zhang, Ping, Zheng, Mengfan, Shi, Yuxuan, Jing, Liwen, Huang, Jianbo, Li, Nan

arXiv.org Artificial Intelligence

Abstract--This paper introduces WirelessGPT, a pioneering foundation model specifically designed for multi-task learning in wireless communication and sensing. In fact, this task-agnostic design adapts WirelessGPT seamlessly to a wide range of downstream tasks, using a unified representation with minimal fine-tuning. By unifying communication and sensing functionalities, WirelessGPT addresses the limitations of task-specific models, offering a scalable and efficient solution for integrated sensing and communication (ISAC). With an initial parameter size of around 80 million, WirelessGPT demonstrates significant improvements over conventional methods and smaller AI models, reducing reliance on large-scale labeled data. As the first foundation model capable of supporting diverse tasks across different domains, WirelessGPT establishes a new benchmark, paving the way for future advancements in multi-task wireless systems.


Strategic Application of AIGC for UAV Trajectory Design: A Channel Knowledge Map Approach

Zhang, Chiya, Wang, Ting, Han, Rubing, Gong, Yuanxiang

arXiv.org Artificial Intelligence

Unmanned Aerial Vehicles (UAVs) are increasingly utilized in wireless communication, yet accurate channel loss prediction remains a significant challenge, limiting resource optimization performance. To address this issue, this paper leverages Artificial Intelligence Generated Content (AIGC) for the efficient construction of Channel Knowledge Maps (CKM) and UAV trajectory design. Given the time-consuming nature of channel data collection, AI techniques are employed in a Wasserstein Generative Adversarial Network (WGAN) to extract environmental features and augment the data. Experiment results demonstrate the effectiveness of the proposed framework in improving CKM construction accuracy. Moreover, integrating CKM into UAV trajectory planning reduces channel gain uncertainty, demonstrating its potential to enhance wireless communication efficiency.


Large Models Enabled Ubiquitous Wireless Sensing

Hu, Shun

arXiv.org Artificial Intelligence

In the era of 5G communication, the knowledge of channel state information (CSI) is crucial for enhancing network performance. This paper explores the utilization of language models for spatial CSI prediction within MIMO-OFDM systems. We begin by outlining the significance of accurate CSI in enabling advanced functionalities such as adaptive modulation. We review existing methodologies for CSI estimation, emphasizing the shift from traditional to data-driven approaches. Then a novel framework for spatial CSI prediction using realistic environment information is proposed, and experimental results demonstrate the effectiveness. This research paves way for innovative strategies in managing wireless networks.


Partial Channel Dependence with Channel Masks for Time Series Foundation Models

Lee, Seunghan, Park, Taeyoung, Lee, Kibok

arXiv.org Machine Learning

Recent advancements in foundation models have been successfully extended to the time series (TS) domain, facilitated by the emergence of large-scale TS datasets. However, previous efforts have primarily focused on designing model architectures to address explicit heterogeneity among datasets such as various numbers of channels, while often overlooking implicit heterogeneity such as varying dependencies between channels. In this work, we introduce the concept of partial channel dependence (PCD), which enables a more sophisticated adjustment of channel dependencies based on dataset-specific information. To achieve PCD, we propose a channel mask that captures the relationships between channels within a dataset using two key components: 1) a correlation matrix that encodes relative dependencies between channels, and 2) domain parameters that learn the absolute dependencies specific to each dataset, refining the correlation matrix. We validate the effectiveness of PCD across four tasks in TS including forecasting, classification, imputation, and anomaly detection, under diverse settings, including few-shot and zero-shot scenarios with both TS foundation models and single-task models. Code is available at https://github.com/seunghan96/CM.


LinFormer: A Linear-based Lightweight Transformer Architecture For Time-Aware MIMO Channel Prediction

Jin, Yanliang, Wu, Yifan, Gao, Yuan, Zhang, Shunqing, Xu, Shugong, Wang, Cheng-Xiang

arXiv.org Artificial Intelligence

The emergence of 6th generation (6G) mobile networks brings new challenges in supporting high-mobility communications, particularly in addressing the issue of channel aging. While existing channel prediction methods offer improved accuracy at the expense of increased computational complexity, limiting their practical application in mobile networks. To address these challenges, we present LinFormer, an innovative channel prediction framework based on a scalable, all-linear, encoder-only Transformer model. Our approach, inspired by natural language processing (NLP) models such as BERT, adapts an encoder-only architecture specifically for channel prediction tasks. We propose replacing the computationally intensive attention mechanism commonly used in Transformers with a time-aware multi-layer perceptron (TMLP), significantly reducing computational demands. The inherent time awareness of TMLP module makes it particularly suitable for channel prediction tasks. We enhance LinFormer's training process by employing a weighted mean squared error loss (WMSELoss) function and data augmentation techniques, leveraging larger, readily available communication datasets. Our approach achieves a substantial reduction in computational complexity while maintaining high prediction accuracy, making it more suitable for deployment in cost-effective base stations (BS). Comprehensive experiments using both simulated and measured data demonstrate that LinFormer outperforms existing methods across various mobility scenarios, offering a promising solution for future wireless communication systems.